2024-07-10 11:55:22.AIbase.10.2k
Ollama 0.2 Release: Default Concurrency for Handling Multiple Requests and Loading Multiple Models Simultaneously
Latest News! Version 0.2 of Ollama has been released! It has been reported that this update enables concurrency by default, allowing Ollama to handle multiple requests simultaneously, delivering a faster user experience. This update not only unlocks the feature of concurrent requests but also supports loading different models simultaneously, making Ollama more efficient in handling various tasks.According to the official news released by Ollama, this update enables Ollama to handle multiple chat